2 research outputs found
CloudSimSC: A Toolkit for Modeling and Simulation of Serverless Computing Environments
Serverless computing is gaining traction as an attractive model for the
deployment of a multitude of workloads in the cloud. Designing and building
effective resource management solutions for any computing environment requires
extensive long term testing, experimentation and analysis of the achieved
performance metrics. Utilizing real test beds and serverless platforms for such
experimentation work is often times not possible due to resource, time and cost
constraints. Thus, employing simulators to model these environments is key to
overcoming the challenge of examining the viability of such novel ideas for
resource management. Existing simulation software developed for serverless
environments lack generalizibility in terms of their architecture as well as
the various aspects of resource management, where most are purely focused on
modeling function performance under a specific platform architecture. In
contrast, we have developed a serverless simulation model with induced
flexibility in its architecture as well as the key resource management aspects
of function scheduling and scaling. Further, we incorporate techniques for
easily deriving monitoring metrics required for evaluating any implemented
solutions by users. Our work is presented as CloudSimSC, a modular extension to
CloudSim which is a simulator tool extensively used for modeling cloud
environments by the research community. We discuss the implemented features in
our simulation tool using multiple use cases
A Deep Reinforcement Learning based Algorithm for Time and Cost Optimized Scaling of Serverless Applications
Serverless computing has gained a strong traction in the cloud computing
community in recent years. Among the many benefits of this novel computing
model, the rapid auto-scaling capability of user applications takes prominence.
However, the offer of adhoc scaling of user deployments at function level
introduces many complications to serverless systems. The added delay and
failures in function request executions caused by the time consumed for
dynamically creating new resources to suit function workloads, known as the
cold-start delay, is one such very prevalent shortcoming. Maintaining idle
resource pools to alleviate this issue often results in wasted resources from
the cloud provider perspective. Existing solutions to address this limitation
mostly focus on predicting and understanding function load levels in order to
proactively create required resources. Although these solutions improve
function performance, the lack of understanding on the overall system
characteristics in making these scaling decisions often leads to the
sub-optimal usage of system resources. Further, the multi-tenant nature of
serverless systems requires a scalable solution adaptable for multiple
co-existing applications, a limitation seen in most current solutions. In this
paper, we introduce a novel multi-agent Deep Reinforcement Learning based
intelligent solution for both horizontal and vertical scaling of function
resources, based on a comprehensive understanding on both function and system
requirements. Our solution elevates function performance reducing cold starts,
while also offering the flexibility for optimizing resource maintenance cost to
the service providers. Experiments conducted considering varying workload
scenarios show improvements of up to 23% and 34% in terms of application
latency and request failures, while also saving up to 45% in infrastructure
cost for the service providers.Comment: 15 pages, 22 figure